Recurrent Neural Learning for Classifying Spoken Utterances
نویسندگان
چکیده
For telecommunications companies or banks, etc processing spontaneous lanaguage in helpdesk scenarios is important for automatic telephone interactions. However, the problem of understanding spontaneous spoken language is difficult. Learning techniques such as neural networks have the ability to learn in a robust manner. Recurrent networks have been used in neurocognitive or psycholinguistically oriented approaches of language processing. Here they are examined for their potential in a difficult spoken language classification task. This paper describes an approach to learning classification of recorded operator assistance telephone utterances. We explore simple recurrent networks using a large, unique telecommunication corpus of spontaneous spoken language. Performance of the network indicates that a simple recurrent network is quite useful for learning classification of spontaneous spoken language in a robust manner, which may lead to their use in helpdesk call routing.
منابع مشابه
Comparing Support Vector Machines, Recurrent Networks, and Finite State Transducers for Classifying Spoken Utterances
This paper describes new experiments for the classification of recorded operator assistance telephone utterances. The experimental work focused on three techniques: support vector machines (SVM), simple recurrent networks (SRN) and finite-state transducers (FST) using a large, unique telecommunication corpus of spontaneous spoken language. A comparison is made of the performance of these classi...
متن کاملA comparison between spiking and differentiable recurrent neural networks on spoken digit recognition
In this paper we demonstrate that Long Short-Term Memory (LSTM) is a differentiable recurrent neural net (RNN) capable of robustly categorizing timewarped speech data. We measure its performance on a spoken digit identification task, where the data was spike-encoded in such a way that classifying the utterances became a difficult challenge in non-linear timewarping. We find that LSTM gives grea...
متن کاملApproximate Inference for Domain Detection in Spoken Language Understanding
This paper presents a semi-latent topic model for semantic domain detection in spoken language understanding systems. We use labeled utterance information to capture latent topics, which directly correspond to semantic domains. Additionally, we introduce an ’informative prior’ for Bayesian inference that can simultaneously segment utterances of known domains into classes and divide them from ou...
متن کاملPredictable and educational spoken dialogues: pilot results
This paper addresses the challenge of designing spoken dialogues that are of educational benefit within the context of an intelligent tutoring system, yet predictable enough to facilitate automatic speech recognition and subsequent processing. We introduce a design principle to meet this goal: construct short dialogues in which the desired student utterances are external evidence of performance...
متن کاملLatticeRnn: Recurrent Neural Networks Over Lattices
We present a new model called LATTICERNN, which generalizes recurrent neural networks (RNNs) to process weighted lattices as input, instead of sequences. A LATTICERNN can encode the complete structure of a lattice into a dense representation, which makes it suitable to a variety of problems, including rescoring, classifying, parsing, or translating lattices using deep neural networks (DNNs). In...
متن کامل